Goto

Collaborating Authors

 average count


FPT: Feature Prompt Tuning for Few-shot Readability Assessment

Wang, Ziyang, Lee, Sanwoo, Huang, Hsiu-Yuan, Wu, Yunfang

arXiv.org Artificial Intelligence

Prompt-based methods have achieved promising results in most few-shot text classification tasks. However, for readability assessment tasks, traditional prompt methods lackcrucial linguistic knowledge, which has already been proven to be essential. Moreover, previous studies on utilizing linguistic features have shown non-robust performance in few-shot settings and may even impair model performance.To address these issues, we propose a novel prompt-based tuning framework that incorporates rich linguistic knowledge, called Feature Prompt Tuning (FPT). Specifically, we extract linguistic features from the text and embed them into trainable soft prompts. Further, we devise a new loss function to calibrate the similarity ranking order between categories. Experimental results demonstrate that our proposed method FTP not only exhibits a significant performance improvement over the prior best prompt-based tuning approaches, but also surpasses the previous leading methods that incorporate linguistic features. Also, our proposed model significantly outperforms the large language model gpt-3.5-turbo-16k in most cases. Our proposed method establishes a new architecture for prompt tuning that sheds light on how linguistic features can be easily adapted to linguistic-related tasks.


Complexity-Guided Curriculum Learning for Text Graphs

Vakil, Nidhi, Amiri, Hadi

arXiv.org Artificial Intelligence

Curriculum learning provides a systematic approach to training. It refines training progressively, tailors training to task requirements, and improves generalization through exposure to diverse examples. We present a curriculum learning approach that builds on existing knowledge about text and graph complexity formalisms for training with text graph data. The core part of our approach is a novel data scheduler, which employs "spaced repetition" and complexity formalisms to guide the training process. We demonstrate the effectiveness of the proposed approach on several text graph tasks and graph neural network architectures. The proposed model gains more and uses less data; consistently prefers text over graph complexity indices throughout training, while the best curricula derived from text and graph complexity indices are equally effective; and it learns transferable curricula across GNN models and datasets. In addition, we find that both node-level (local) and graph-level (global) graph complexity indices, as well as shallow and traditional text complexity indices play a crucial role in effective curriculum learning.


Inventory Demand Forecasting using Machine Learning - Python - GeeksforGeeks

#artificialintelligence

The vendors who are selling everyday items need to keep their stock up to date so, that no customer returns from their shop empty hand. In this article, we will try to implement a machine learning model which can predict the stock amount for the different products which are sold in different stores. Python libraries make it easy for us to handle the data and perform typical and complex tasks with a single line of code. Now let's load the dataset into the panda's data frame and print its first five rows. Now let's check the size we have calculated is correct or not .


Pushing on Text Readability Assessment: A Transformer Meets Handcrafted Linguistic Features

Lee, Bruce W., Jang, Yoo Sung, Lee, Jason Hyung-Jong

arXiv.org Artificial Intelligence

We report two essential improvements in readability assessment: 1. three novel features in advanced semantics and 2. the timely evidence that traditional ML models (e.g. Random Forest, using handcrafted features) can combine with transformers (e.g. RoBERTa) to augment model performance. First, we explore suitable transformers and traditional ML models. Then, we extract 255 handcrafted linguistic features using self-developed extraction software. Finally, we assemble those to create several hybrid models, achieving state-of-the-art (SOTA) accuracy on popular datasets in readability assessment. The use of handcrafted features help model performance on smaller datasets. Notably, our RoBERTA-RF-T1 hybrid achieves the near-perfect classification accuracy of 99%, a 20.3% increase from the previous SOTA.